Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 17.245
Filtrar
1.
J Acoust Soc Am ; 155(4): 2627-2635, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38629884

RESUMO

Passive acoustic monitoring (PAM) is an optimal method for detecting and monitoring cetaceans as they frequently produce sound while underwater. Cue counting, counting acoustic cues of deep-diving cetaceans instead of animals, is an alternative method for density estimation, but requires an average cue production rate to convert cue density to animal density. Limited information about click rates exists for sperm whales in the central North Pacific Ocean. In the absence of acoustic tag data, we used towed hydrophone array data to calculate the first sperm whale click rates from this region and examined their variability based on click type, location, distance of whales from the array, and group size estimated by visual observers. Our findings show click type to be the most important variable, with groups that include codas yielding the highest click rates. We also found a positive relationship between group size and click detection rates that may be useful for acoustic predictions of group size in future studies. Echolocation clicks detected using PAM methods are often the only indicator of deep-diving cetacean presence. Understanding the factors affecting their click rates provides important information for acoustic density estimation.


Assuntos
Ecolocação , Cachalote , Animais , Vocalização Animal , Acústica , Baleias , Espectrografia do Som
2.
J Acoust Soc Am ; 155(4): 2612-2626, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38629882

RESUMO

This study presents an acoustic investigation of the vowel inventory of Drehu (Southern Oceanic Linkage), spoken in New Caledonia. Reportedly, Drehu has a 14 vowel system distinguishing seven vowel qualities and an additional length distinction. Previous phonological descriptions were based on impressionistic accounts showing divergent proposals for two out of seven reported vowel qualities. This study presents the first phonetic investigation of Drehu vowels based on acoustic data from eight speakers. To examine the phonetic correlates of the proposed phonological vowel inventory, multi-point acoustic analyses were used, and vowel inherent spectral change (VISC) was investigated (F1, F2, and F3). Additionally, vowel duration was measured. Contrary to reports from other studies on VISC in monophthongs, we find that monophthongs in Drehu are mostly steady state. We propose a revised vowel inventory and focus on the acoustic description of open-mid /ɛ/ and the central vowel /ə/, whose status was previously unclear. Additionally, we find that vowel quality stands orthogonal to vowel quantity by demonstrating that the phonological vowel length distinction is primarily based on a duration cue rather than formant structure. Finally, we report the acoustic properties of the seven vowel qualities that were identified.


Assuntos
Fonética , Acústica da Fala , Acústica
3.
J Texture Stud ; 55(2): e12832, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38613251

RESUMO

Puffed-grain food is a crispy snack whose consumer satisfaction depends on snack crispness and crunchiness, which can be characterized by the sound and the acoustic signals of food breaking. This study aimed to evaluate whether acoustic characteristics can be used to predict the crispness of various puffed-grain food. Sensory evaluation was performed on puffed-grain products with varying hygroscopic durations and different types. The relation between sensory evaluation and acoustic characteristics of nine different types of food was examined. The Hilbert-Huang transform was used to perform energy segmentation of the acoustic signal of puffed-grain food and observe its energy migration process. The results showed that energy release was more concentrated in the low-frequency range for grain-puffed foods with different hygroscopic durations. No notable correlation was observed between the low-frequency interval and sensory crispness for the different types of puffed-grain foods. However, the acoustic features extracted from their inherent low-frequency intervals showed a significantly improved correlation with sensory crispness. Therefore, it provides a theoretical reference for applying acoustic characteristics to describe food texture.


Assuntos
Acústica , Som , Grão Comestível , Fenômenos Físicos , Lanches
4.
Sensors (Basel) ; 24(7)2024 Mar 22.
Artigo em Inglês | MEDLINE | ID: mdl-38610256

RESUMO

The ongoing biodiversity crisis, driven by factors such as land-use change and global warming, emphasizes the need for effective ecological monitoring methods. Acoustic monitoring of biodiversity has emerged as an important monitoring tool. Detecting human voices in soundscape monitoring projects is useful both for analyzing human disturbance and for privacy filtering. Despite significant strides in deep learning in recent years, the deployment of large neural networks on compact devices poses challenges due to memory and latency constraints. Our approach focuses on leveraging knowledge distillation techniques to design efficient, lightweight student models for speech detection in bioacoustics. In particular, we employed the MobileNetV3-Small-Pi model to create compact yet effective student architectures to compare against the larger EcoVAD teacher model, a well-regarded voice detection architecture in eco-acoustic monitoring. The comparative analysis included examining various configurations of the MobileNetV3-Small-Pi-derived student models to identify optimal performance. Additionally, a thorough evaluation of different distillation techniques was conducted to ascertain the most effective method for model selection. Our findings revealed that the distilled models exhibited comparable performance to the EcoVAD teacher model, indicating a promising approach to overcoming computational barriers for real-time ecological monitoring.


Assuntos
Fala , Voz , Humanos , Acústica , Biodiversidade , Conhecimento
5.
Sensors (Basel) ; 24(7)2024 Mar 29.
Artigo em Inglês | MEDLINE | ID: mdl-38610400

RESUMO

Monitoring blood pressure, a parameter closely related to cardiovascular activity, can help predict imminent cardiovascular events. In this paper, a novel method is proposed to customize an existing mechanistic model of the cardiovascular system through feature extraction from cardiopulmonary acoustic signals to estimate blood pressure using artificial intelligence. As various factors, such as drug consumption, can alter the biomechanical properties of the cardiovascular system, the proposed method seeks to personalize the mechanistic model using information extracted from vibroacoustic sensors. Simulation results for the proposed approach are evaluated by calculating the error in blood pressure estimates compared to ground truth arterial line measurements, with the results showing promise for this method.


Assuntos
Inteligência Artificial , Sistema Cardiovascular , Pressão Sanguínea , Determinação da Pressão Arterial , Acústica
6.
Sensors (Basel) ; 24(7)2024 Apr 04.
Artigo em Inglês | MEDLINE | ID: mdl-38610510

RESUMO

The perception of sound greatly impacts users' emotional states, expectations, affective relationships with products, and purchase decisions. Consequently, assessing the perceived quality of sounds through jury testing is crucial in product design. However, the subjective nature of jurors' responses may limit the accuracy and reliability of jury test outcomes. This research explores the utility of facial expression analysis in jury testing to enhance response reliability and mitigate subjectivity. Some quantitative indicators allow the research hypothesis to be validated, such as the correlation between jurors' emotional responses and valence values, the accuracy of jury tests, and the disparities between jurors' questionnaire responses and the emotions measured by FER (facial expression recognition). Specifically, analysis of attention levels during different statuses reveals a discernible decrease in attention levels, with 70 percent of jurors exhibiting reduced attention levels in the 'distracted' state and 62 percent in the 'heavy-eyed' state. On the other hand, regression analysis shows that the correlation between jurors' valence and their choices in the jury test increases when considering the data where the jurors are attentive. The correlation highlights the potential of facial expression analysis as a reliable tool for assessing juror engagement. The findings suggest that integrating facial expression recognition can enhance the accuracy of jury testing in product design by providing a more dependable assessment of user responses and deeper insights into participants' reactions to auditory stimuli.


Assuntos
Reconhecimento Facial , Humanos , Reprodutibilidade dos Testes , Acústica , Som , Emoções
7.
Noise Health ; 26(120): 1-7, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38570303

RESUMO

OBJECTIVE: Functional dysphonia can impair the language expression ability and adversely affect the career development of some patients. Therefore, an active exploration of effective treatment options is imperative. This study investigated the effect of Akson therapy on acoustic parameters in patients with functional dysphonia. MATERIALS AND METHODS: In this retrospective analysis, 79 patients with functional dysphonia who received conventional voice correction training from June 2020 to June 2021 were included in the reference group (RG). Our hospital has implemented Akson therapy since July 2021. Correspondingly, 72 patients with functional dysphonia who underwent Akson therapy from July 2021 to July 2022 were enrolled in the observation group (OG). The acoustic parameters such as fundamental frequency (F0), jitter, shimmer, and normalized noise energy (NNE); the aerodynamic parameters including maximum phonation time (MPT), mean airflow rate (MFR), and Voice Handicap Index-10 (VHI-10) score; and the Grade, Roughness, Breathiness, Asthenia, and Strain scale (GRBAS) score were measured before and after treatment and compared between the two groups. RESULTS: The F0, jitter, shimmer, NNE, MPT, and MFR values as well as the VHI-10 score and the grade (G), roughness (R), and breathiness (B) scores on the GRBAS did not significantly differ between the two groups before treatment (P > 0.05). However, significantly lower F0, jitter, shimmer, NNE, and MFR values and higher MPT levels were found in the OG compared to the RG after treatment (P < 0.001). Furthermore, the VHI-10 score and the G, R, and B scores were significantly lower in the OG than in the RG after treatment (P < 0.001), whereas the asthenia (A) and strain (S) scores remained at 0 before and after treatment. CONCLUSION: Akson therapy can improve the acoustic parameters of patients with functional dysphonia to a certain extent, indicating its potential application value.


Assuntos
Disfonia , Humanos , Disfonia/terapia , Estudos Retrospectivos , Astenia , Qualidade da Voz , Acústica
8.
JASA Express Lett ; 4(4)2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38568027

RESUMO

This study investigates speech production under various room acoustic conditions in virtual environments, by comparing vocal behavior and the subjective experience of speaking in four real rooms and their audio-visual virtual replicas. Sex differences were explored. Males and females (N = 13) adjusted their voice levels similarly to room acoustic changes in the real rooms, but only males did so in the virtual rooms. Females, however, rated the visual virtual environment as more realistic compared to males. This suggests a discrepancy between sexes regarding the experience of realism in a virtual environment and changes in objective behavioral measures such as voice level.


Assuntos
Caracteres Sexuais , Fala , Feminino , Masculino , Humanos , Acústica
9.
J Acoust Soc Am ; 155(4): 2285-2301, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38557735

RESUMO

Fronting of the vowels /u, ʊ, o/ is observed throughout most North American English varieties, but has been analyzed mainly in terms of acoustics rather than articulation. Because an increase in F2, the acoustic correlate of vowel fronting, can be the result of any gesture that shortens the front cavity of the vocal tract, acoustic data alone do not reveal the combination of tongue fronting and/or lip unrounding that speakers use to produce fronted vowels. It is furthermore unresolved to what extent the articulation of fronted back vowels varies according to consonantal context and how the tongue and lips contribute to the F2 trajectory throughout the vowel. This paper presents articulatory and acoustic data on fronted back vowels from two varieties of American English: coastal Southern California and South Carolina. Through analysis of dynamic acoustic, ultrasound, and lip video data, it is shown that speakers of both varieties produce fronted /u, ʊ, o/ with rounded lips, and that high F2 observed for these vowels is associated with a front-central tongue position rather than unrounded lips. Examination of time-varying formant trajectories and articulatory configurations shows that the degree of vowel-internal F2 change is predominantly determined by coarticulatory influence of the coda.


Assuntos
Fonética , Acústica da Fala , Estados Unidos , Acústica , Idioma , South Carolina
10.
J Acoust Soc Am ; 155(4): R7-R8, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38558083

RESUMO

The Reflections series takes a look back on historical articles from The Journal of the Acoustical Society of America that have had a significant impact on the science and practice of acoustics.


Assuntos
Percepção da Fala , Acústica , Acústica da Fala , Cognição
11.
JASA Express Lett ; 4(4)2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38563690

RESUMO

Moose are a popular species with recreationists but understudied acoustically. We used publicly available videos to characterize and quantify the vocalizations of moose in New Hampshire separated by age/sex class. We found significant differences in peak frequency, center frequency, bandwidth, and duration across the groups. Our results provide quantification of wild moose vocalizations across age/sex classes, which is a key step for passive acoustic detection of this species and highlights public videos as a potential resource for bioacoustics research of hard-to-capture and understudied species.


Assuntos
Cervos , Animais , Acústica , New Hampshire , Gravação de Videoteipe
12.
J Acoust Soc Am ; 155(4): 2577-2588, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38602448

RESUMO

Classifying acoustic responses captured through earphones offers valuable insights into nearby environments, such as whether the earphones are in or out of the ear. However, the performances of classification algorithms often suffer when applied to other devices due to domain mismatches. This study proposes a domain-adaptation method tailored for acoustic-response data from two distinct insert earphone models. The method trains a domain-adaptation function using a pair of datasets obtained from a set of acoustic loads, yielding a domain-adapted dataset suitable for training classification algorithms in a target domain. The effectiveness of this approach is validated through assessments of domain adaptation quality and resulting performance enhancements in the classification algorithm tasked with discerning whether an earphone is positioned inside or outside the ear. Importantly, our method requires significantly fewer measurements than the original dataset, reducing data collection time while providing a suitable training dataset for the target domain. Additionally, the method's reusability across future devices streamlines data collection time and efforts for the future devices.


Assuntos
Acústica , Algoritmos
13.
J Acoust Soc Am ; 155(4): 2538-2548, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38591939

RESUMO

Long-term fixed passive acoustic monitoring of cetacean populations is a logistical and technological challenge, often limited by the battery capacity of the autonomous recorders. Depending on the research scope and target species, temporal subsampling of the data may become necessary to extend the deployment period. This study explores the effects of different duty cycles on metrics that describe patterns of seasonal presence, call type richness richness, and daily call rate of three blue whale acoustics populations in the Southern Indian Ocean. Detections of blue whale calls from continuous acoustic data were subsampled with three different duty cycles of 50%, 33%, and 25% within listening periods ranging from 1 min to 6 h. Results show that reducing the percentage of recording time reduces the accuracy of the observed seasonal patterns as well as the estimation of daily call rate and call call type richness. For a specific duty cycle, short listening periods (5-30 min) are preferred to longer listening periods (1-6 h). The effects of subsampling are greater the lower the species' vocal activity or the shorter their periods of presence. These results emphasize the importance of selecting a subsampling scheme adapted to the target species.


Assuntos
Acústica , Balaenoptera , Animais , Cetáceos , Fontes de Energia Elétrica , Oceano Índico
14.
PLoS One ; 19(4): e0301514, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38564597

RESUMO

Evoked potential studies have shown that speech planning modulates auditory cortical responses. The phenomenon's functional relevance is unknown. We tested whether, during this time window of cortical auditory modulation, there is an effect on speakers' perceptual sensitivity for vowel formant discrimination. Participants made same/different judgments for pairs of stimuli consisting of a pre-recorded, self-produced vowel and a formant-shifted version of the same production. Stimuli were presented prior to a "go" signal for speaking, prior to passive listening, and during silent reading. The formant discrimination stimulus /uh/ was tested with a congruent productions list (words with /uh/) and an incongruent productions list (words without /uh/). Logistic curves were fitted to participants' responses, and the just-noticeable difference (JND) served as a measure of discrimination sensitivity. We found a statistically significant effect of condition (worst discrimination before speaking) without congruency effect. Post-hoc pairwise comparisons revealed that JND was significantly greater before speaking than during silent reading. Thus, formant discrimination sensitivity was reduced during speech planning regardless of the congruence between discrimination stimulus and predicted acoustic consequences of the planned speech movements. This finding may inform ongoing efforts to determine the functional relevance of the previously reported modulation of auditory processing during speech planning.


Assuntos
Córtex Auditivo , Percepção da Fala , Humanos , Fala/fisiologia , Percepção da Fala/fisiologia , Acústica , Movimento , Fonética , Acústica da Fala
15.
Am Nat ; 203(5): 618-627, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-38635364

RESUMO

AbstractAutonomous sensors provide opportunities to observe organisms across spatial and temporal scales that humans cannot directly observe. By processing large data streams from autonomous sensors with deep learning methods, researchers can make novel and important natural history discoveries. In this study, we combine automated acoustic monitoring with deep learning models to observe breeding-associated activity in the endangered Sierra Nevada yellow-legged frog (Rana sierrae), a behavior that current surveys do not measure. By deploying inexpensive hydrophones and developing a deep learning model to recognize breeding-associated vocalizations, we discover three undocumented R. sierrae vocalization types and find an unexpected temporal pattern of nocturnal breeding-associated vocal activity. This study exemplifies how the combination of autonomous sensor data and deep learning can shed new light on species' natural history, especially during times or in locations where human observation is limited or impossible.


Assuntos
Ranidae , Vocalização Animal , Animais , Humanos , Acústica
16.
J Acoust Soc Am ; 155(4): 2603-2611, 2024 Apr 01.
Artigo em Inglês | MEDLINE | ID: mdl-38629881

RESUMO

Open science practices have led to an increase in available speech datasets for researchers interested in acoustic analysis. Accurate evaluation of these databases frequently requires manual or semi-automated analysis. The time-intensive nature of these analyses makes them ideally suited for research assistants in laboratories focused on speech and voice production. However, the completion of high-quality, consistent, and reliable analyses requires clear rules and guidelines for all research assistants to follow. This tutorial will provide information on training and mentoring research assistants to complete these analyses, covering areas including RA training, ongoing data analysis monitoring, and documentation needed for reliable and re-creatable findings.


Assuntos
Distúrbios da Voz , Voz , Humanos , Acústica , Fala
17.
Elife ; 132024 Apr 18.
Artigo em Inglês | MEDLINE | ID: mdl-38635312

RESUMO

Complex skills like speech and dance are composed of ordered sequences of simpler elements, but the neuronal basis for the syntactic ordering of actions is poorly understood. Birdsong is a learned vocal behavior composed of syntactically ordered syllables, controlled in part by the songbird premotor nucleus HVC (proper name). Here, we test whether one of HVC's recurrent inputs, mMAN (medial magnocellular nucleus of the anterior nidopallium), contributes to sequencing in adult male Bengalese finches (Lonchura striata domestica). Bengalese finch song includes several patterns: (1) chunks, comprising stereotyped syllable sequences; (2) branch points, where a given syllable can be followed probabilistically by multiple syllables; and (3) repeat phrases, where individual syllables are repeated variable numbers of times. We found that following bilateral lesions of mMAN, acoustic structure of syllables remained largely intact, but sequencing became more variable, as evidenced by 'breaks' in previously stereotyped chunks, increased uncertainty at branch points, and increased variability in repeat numbers. Our results show that mMAN contributes to the variable sequencing of vocal elements in Bengalese finch song and demonstrate the influence of recurrent projections to HVC. Furthermore, they highlight the utility of species with complex syntax in investigating neuronal control of ordered sequences.


Assuntos
Aves Canoras , Masculino , Animais , Fala , Acústica , Memória , Comportamento Estereotipado
18.
Sci Rep ; 14(1): 8977, 2024 Apr 18.
Artigo em Inglês | MEDLINE | ID: mdl-38637516

RESUMO

Why do we prefer some singers to others? We investigated how much singing voice preferences can be traced back to objective features of the stimuli. To do so, we asked participants to rate short excerpts of singing performances in terms of how much they liked them as well as in terms of 10 perceptual attributes (e.g.: pitch accuracy, tempo, breathiness). We modeled liking ratings based on these perceptual ratings, as well as based on acoustic features and low-level features derived from Music Information Retrieval (MIR). Mean liking ratings for each stimulus were highly correlated between Experiments 1 (online, US-based participants) and 2 (in the lab, German participants), suggesting a role for attributes of the stimuli in grounding average preferences. We show that acoustic and MIR features barely explain any variance in liking ratings; in contrast, perceptual features of the voices achieved around 43% of prediction. Inter-rater agreement in liking and perceptual ratings was low, indicating substantial (and unsurprising) individual differences in participants' preferences and perception of the stimuli. Our results indicate that singing voice preferences are not grounded in acoustic attributes of the voices per se, but in how these features are perceptually interpreted by listeners.


Assuntos
Música , Canto , Voz , Humanos , Qualidade da Voz , Acústica
19.
Eur Rev Med Pharmacol Sci ; 28(7): 2701-2709, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38639510

RESUMO

OBJECTIVE: Vocal cord paralysis (VCP) is a serious complication in thyroidectomy operations; however, its management remains unclear. The present study evaluated the voice parameters of patients who underwent surgery using Intraoperative Neurophysiologic Monitoring (IONM). PATIENTS AND METHODS: A total of 52 patients (41 females and 11 males) who underwent a total thyroidectomy operation were evaluated using objective and subjective voice analysis examinations before and after surgery. Acoustic parameters, such as Fundamental Frequency (F0), Shimmer, Jitter, Noise-to-Harmonic ratio (NHR), and aerodynamic parameters, including S/Z ratio and maximum phonation time (MPT), were analyzed. Objective findings, including the VHI-10 (Voice Handicap Index) and V-RQOL (Voice-Related Quality of Life), were also analyzed. The relationship between voice parameters and IONM values was investigated. RESULTS: The objective analysis (acoustic and aerodynamic parameters) showed no difference (p>0.05). However, the subjective analysis, which involved the VHI-10 and V-RQOL measures, revealed a significant difference before and after the operation (p<0.05). The Spearman correlation analysis showed that the NHR postoperative 1st-month parameter negatively correlated (rho=-0.317, p<0.059), while the F0 postoperative 6th-month parameter positively correlated (rho=0.347) with the amplitude difference before and after dissection (Right R2-R1 difference) for the right RLN measured in IONM. CONCLUSIONS: Patients who are planning to undergo a thyroidectomy procedure should undergo voice assessment during both the preoperative and postoperative periods. IONM could improve voice quality outcomes.


Assuntos
Paralisia das Pregas Vocais , Distúrbios da Voz , Masculino , Feminino , Humanos , Qualidade da Voz , Tireoidectomia/efeitos adversos , Qualidade de Vida , Acústica , Paralisia das Pregas Vocais/diagnóstico , Paralisia das Pregas Vocais/etiologia , Distúrbios da Voz/diagnóstico , Distúrbios da Voz/etiologia
20.
Sci Rep ; 14(1): 8748, 2024 04 16.
Artigo em Inglês | MEDLINE | ID: mdl-38627566

RESUMO

Efficient techniques for separating target cells from undiluted blood are necessary for various diagnostic and research applications. This paper presents acoustic focusing in dense media containing iodixanol to purify peripheral blood mononuclear cells (PBMCs) from whole blood in a label-free and flow-through format. If the blood is laminated or mixed with iodixanol solutions while passing through the resonant microchannel, all the components (fluids and cells) rearrange according to their acoustic impedances. Red blood cells (RBCs) have higher effective acoustic impedance than PBMCs. Therefore, they relocate to the pressure node despite the dense medium, while PBMCs stay near the channel walls due to their negative contrast factor relative to their surrounding medium. By modifying the medium and thus tuning the contrast factor of the cells, we enriched PBMCs relative to RBCs by a factor of 3600 to 11,000 and with a separation efficiency of 85%. That level of RBC depletion is higher than most other microfluidic methods and similar to that of density gradient centrifugation. The current acoustophoretic chip runs up to 20 µl/min undiluted whole blood and can be integrated with downstream analysis.


Assuntos
Leucócitos Mononucleares , Técnicas Analíticas Microfluídicas , Separação Celular/métodos , Ácidos Tri-Iodobenzoicos , Acústica , Técnicas Analíticas Microfluídicas/métodos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...